1,348 research outputs found
Evaluation Measures for Relevance and Credibility in Ranked Lists
Recent discussions on alternative facts, fake news, and post truth politics
have motivated research on creating technologies that allow people not only to
access information, but also to assess the credibility of the information
presented to them by information retrieval systems. Whereas technology is in
place for filtering information according to relevance and/or credibility, no
single measure currently exists for evaluating the accuracy or precision (and
more generally effectiveness) of both the relevance and the credibility of
retrieved results. One obvious way of doing so is to measure relevance and
credibility effectiveness separately, and then consolidate the two measures
into one. There at least two problems with such an approach: (I) it is not
certain that the same criteria are applied to the evaluation of both relevance
and credibility (and applying different criteria introduces bias to the
evaluation); (II) many more and richer measures exist for assessing relevance
effectiveness than for assessing credibility effectiveness (hence risking
further bias).
Motivated by the above, we present two novel types of evaluation measures
that are designed to measure the effectiveness of both relevance and
credibility in ranked lists of retrieval results. Experimental evaluation on a
small human-annotated dataset (that we make freely available to the research
community) shows that our measures are expressive and intuitive in their
interpretation
Rich Ground State Chemical Ordering in Nanoparticles: Exact Solution of a Model for Ag-Au Clusters
We show that nanoparticles can have very rich ground state chemical order.
This is illustrated by determining the chemical ordering of Ag-Au 309-atom
Mackay icosahedral nanoparticles. The energy of the nanoparticles is described
using a cluster expansion model, and a Mixed Integer Programming (MIP) approach
is used to find the exact ground state configurations for all stoichiometries.
The chemical ordering varies widely between the different stoichiometries, and
display a rich zoo of structures with non-trivial ordering.Comment: Revised version. New figure added, discussion expanded, some material
moved into supplementary fil
Dual-labelled (13C/15N) green manure to differentiate between plant uptake of organic and inorganic N
Experimental data is still lacking for determining whether plant uptake of organic nitrogen in agricultural soils contributes substantially to the total N uptake. Pulseinjection studies with dual-labelled amino acids have confirmed that non-mycorrhizal crops possess the capacity to take up organic N but failed to quantify the uptake relative to total N uptake
Deep Learning Relevance: Creating Relevant Information (as Opposed to Retrieving it)
What if Information Retrieval (IR) systems did not just retrieve relevant
information that is stored in their indices, but could also "understand" it and
synthesise it into a single document? We present a preliminary study that makes
a first step towards answering this question. Given a query, we train a
Recurrent Neural Network (RNN) on existing relevant information to that query.
We then use the RNN to "deep learn" a single, synthetic, and we assume,
relevant document for that query. We design a crowdsourcing experiment to
assess how relevant the "deep learned" document is, compared to existing
relevant documents. Users are shown a query and four wordclouds (of three
existing relevant documents and our deep learned synthetic document). The
synthetic document is ranked on average most relevant of all.Comment: Neu-IR '16 SIGIR Workshop on Neural Information Retrieval, July 21,
2016, Pisa, Ital
Entropy and Graph Based Modelling of Document Coherence using Discourse Entities: An Application
We present two novel models of document coherence and their application to
information retrieval (IR). Both models approximate document coherence using
discourse entities, e.g. the subject or object of a sentence. Our first model
views text as a Markov process generating sequences of discourse entities
(entity n-grams); we use the entropy of these entity n-grams to approximate the
rate at which new information appears in text, reasoning that as more new words
appear, the topic increasingly drifts and text coherence decreases. Our second
model extends the work of Guinaudeau & Strube [28] that represents text as a
graph of discourse entities, linked by different relations, such as their
distance or adjacency in text. We use several graph topology metrics to
approximate different aspects of the discourse flow that can indicate
coherence, such as the average clustering or betweenness of discourse entities
in text. Experiments with several instantiations of these models show that: (i)
our models perform on a par with two other well-known models of text coherence
even without any parameter tuning, and (ii) reranking retrieval results
according to their coherence scores gives notable performance gains, confirming
a relation between document coherence and relevance. This work contributes two
novel models of document coherence, the application of which to IR complements
recent work in the integration of document cohesiveness or comprehensibility to
ranking [5, 56]
Collembola’s role in regulating mass fluxes in soil and the effects of contrasting life histories
The study determined metabolic rates and elemental pools for two Collembola species with contrasting life histories. The fittest of the two species, P. minuta, excreted the equivalent of 10–12% of the elemental body content per day, and P. armata 7–10%. Most elements are lost to excretion (CO2 and N-waste). These figures in combination with stoichiometry and life histories indicate that the cost of P. minuta’s better fitness is a requirement for a higher quality diet than P. armata. The data produced in this study can be used to estimate the collembolan contribution to C and N fluxes in the soil
SensibleSleep: A Bayesian Model for Learning Sleep Patterns from Smartphone Events
We propose a Bayesian model for extracting sleep patterns from smartphone
events. Our method is able to identify individuals' daily sleep periods and
their evolution over time, and provides an estimation of the probability of
sleep and wake transitions. The model is fitted to more than 400 participants
from two different datasets, and we verify the results against ground truth
from dedicated armband sleep trackers. We show that the model is able to
produce reliable sleep estimates with an accuracy of 0.89, both at the
individual and at the collective level. Moreover the Bayesian model is able to
quantify uncertainty and encode prior knowledge about sleep patterns. Compared
with existing smartphone-based systems, our method requires only screen on/off
events, and is therefore much less intrusive in terms of privacy and more
battery-efficient
- …